14 - Artificial Intelligence II [ID:52586]
50 von 729 angezeigt

One person who shows up at the last minute for some reason.

I'm wondering whether that's like an artifact of the system or whether that's an actual

person.

But always at like 24 it goes up by one.

Anyway, right before we start one reminder again that there is going to be this teaching

assessment thingy the day after tomorrow.

I put all the details on Stutt on so you can read up on them there.

Someone else suggested that since that basically removes half of the lecture on Thursday anyway

we might use the remainder of the session for something like a Q&A.

I would be happy to do that.

I will prepare slides anyway because I'm not sure whether we're going to manage filling

an entire remaining half of the lecture with questions alone.

I'd be delighted to be proven wrong.

But just in case if you have any questions if there's anything you would want me to

repeat with respect to the subject matter of the semester.

Prepare questions think about something or whatever we can gladly use the rest of the

session to like try to sort those things out before we continue with new stuff.

So last time we started with talking about machine learning in particular supervised

learning.

The general set up in the case of supervised learning is that we assume we have some kind

of function f from some domain A to some codomain B which we don't know.

And we have a bunch of hypotheses all of which are basically functions from A to B and we

have a set of examples which you can think of as a function on a subset of A to B or

IE a function on the subset of the domain.

We call those the examples or the training set and the goal is to derive some hypothesis

in H that is ideally consistent with all of the examples that we have with the goal of

that approximating the true function f which we don't actually know as well as by any means

possible.

In practice it's not always the case that our training set actually does represent a

function there is usually noise in there, there might be conflicting classification

these kinds of things.

But in order to like derive all of the like general techniques for machine learning we

basically just assume it is because that makes the problem easier and then some training

methods are more robust against the case where there are conflicting labels than others and

all these kinds of things.

If our codomain is just some finite set we call it a classification problem, if the codomain

is the reals we call it a regression problem.

Obviously we should to some extent try to make sure that the functions that we're trying

to learn or our hypothesis base is as simple as possible and of course as complicated as

necessary.

If we don't we get these kinds of artifacts which we're going to call over fitting later

where it turns out that we find some hypothesis that actually perfectly matches all of the

data that we have all of our training examples but it's just overly complicated with the

result that it's probably not going to generalize as well as simpler hypotheses tend to do.

So in generally keep the hypothesis base as small as you can without losing realizability

i.e. the possibility that our hypothesis base contains the function that we're trying to

learn in the first place.

In general we assume that all of the data that we have is IID independent and identically

distributed.

That is mostly an assumption that we use in order to derive all of these machine learning

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

01:19:40 Min

Aufnahmedatum

2024-06-11

Hochgeladen am

2024-06-13 02:09:11

Sprache

en-US

Einbetten
Wordpress FAU Plugin
iFrame
Teilen